A Day in the Life of a Data Cache Miss

نویسنده

  • Tejas Karkhanis
چکیده

The activity within a processor following a cache miss is studied via a series of simulation experiments. This is a preliminary step toward developing ways of mitigating data cache miss penalties, especially for long misses. With a modest-sized reorder buffer (ROB) of 64 entries, structural blockages due to a full ROB are the major cause of the cache miss penalty. For the SpecINT2000 benchmarks, about 90% of long cache misses result in a blocked ROB. After structural constraints are removed, data dependences are found not to be a major cause of performance loss. In fact, instruction issuing can proceed at full speed for at least one thousand cycles (the length we sampled) beyond a data cache miss. In some cases we found control dependences do pose a problem, however. That is, there are many mispredicted branches that depend on data from a cache miss. With an 8K gshare predictor, 30% or more of data cache misses feed into a mispredicted branch in 7 of 12 benchmarks. Finally, we discuss implications for future processor designs. This includes lengthening ROBs (and enlarging physical rename locations), coupled with a more modest enlarging of instruction issue windows.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Enhanced Flush+Reload Attack on AES

In cloud computing, multiple users can share the same physical machine that can potentially leak secret information, in particular when the memory de-duplication is enabled. Flush+Reload attack is a cache-based attack that makes use of resource sharing. T-table implementation of AES is commonly used in the crypto libraries like OpenSSL. Several Flush+Reload attacks on T-table implementat...

متن کامل

High-level Data-Layout Transformations for Energy Reduction in Direct Mapped Caches

In order to reduce power consumption in data dominated embedded applications, a multi-level memory hierarchy containing caches is often introduced in the implementation. The cache memory used in this hierarchy must have simple tag-comparison mechanisms and tag array overhead for reduced energy. However, such cache architecture, a direct mapped cache, potentially also causes energy overhead due ...

متن کامل

Reduction in Cache Memory Power Consumption based on Replacement Quantity

Today power consumption is considered to be one of the important issues. Therefore, its reduction plays a considerable role in developing systems. Previous studies have shown that approximately 50% of total power consumption is used in cache memories. There is a direct relationship between power consumption and replacement quantity made in cache. The less the number of replacements is, the less...

متن کامل

Reduction in Cache Memory Power Consumption based on Replacement Quantity

Today power consumption is considered to be one of the important issues. Therefore, its reduction plays a considerable role in developing systems. Previous studies have shown that approximately 50% of total power consumption is used in cache memories. There is a direct relationship between power consumption and replacement quantity made in cache. The less the number of replacements is, the less...

متن کامل

Predicting Load Latencies Using Cache Profiling

cache profiling, simulation, cache management instructions, prefetching, memory system analysis, program characterization Due to increasing cache-miss latencies, cache control instructions are being implemented for future systems. In order to investigate the potential benefit of using these instructions in compiling a broad range of applications, we study the memory referencing behavior of indi...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2002